Submodular optimization problems and greedy strategies: A survey
نویسندگان
چکیده
منابع مشابه
Sparse and Greedy: Sparsifying Submodular Facility Location Problems
Submodular facility location functions are widely used for summarizing large datasets and have found applications ranging from sensor placement, image retrieval, and clustering. A significant problem is that evaluating such functions typically requires the calculation of pairwise benefits for all items, which is computationally unmanageable for large problems. In this paper we propose a sparsif...
متن کاملGreedy Strategies for Convex Optimization
We investigate two greedy strategies for finding an approximation to the minimum of a convex function E defined on a Hilbert space H. We prove convergence rates for these algorithms under suitable conditions on the objective function E. These conditions involve the behavior of the modulus of smoothness and the modulus of uniform convexity of E.
متن کاملGreed Is Good: Near-Optimal Submodular Maximization via Greedy Optimization
It is known that greedy methods perform well for maximizing monotone submodular functions. At the same time, such methods perform poorly in the face of non-monotonicity. In this paper, we show—arguably, surprisingly—that invoking the classical greedy algorithm O( √ k)-times leads to the (currently) fastest deterministic algorithm, called REPEATEDGREEDY, for maximizing a general submodular funct...
متن کاملParallel Double Greedy Submodular Maximization
Many machine learning problems can be reduced to the maximization of submodular functions. Although well understood in the serial setting, the parallel maximization of submodular functions remains an open area of research with recent results [1] only addressing monotone functions. The optimal algorithm for maximizing the more general class of non-monotone submodular functions was introduced by ...
متن کاملGreedy Maximization of Submodular Functions
Traditional optimization techniques often rely upon functions that are convex or at least locally convex. Such diverse methods as gradient descent, loopy belief propagation, and linear programming all rely upon convex functions. However, many natural functions are not convex, yet optimizing over them is both possible and necessary. The class of submodular functions is particularly well-behaved ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Discrete Event Dynamic Systems
سال: 2020
ISSN: 0924-6703,1573-7594
DOI: 10.1007/s10626-019-00308-7